Second Order Derivatives for Network Pruning: Optimal Brain Surgeon

نویسندگان

  • Babak Hassibi
  • David G. Stork
چکیده

We investigate the use of information from all second order derivatives of the error function to perform network pruning (i.e., removing unimportant weights from a trained network) in order to improve generalization and increase the speed of further training. Our method, Optimal Brain Surgeon (OBS), is significantly better than magnitude-based methods, which can often remove the wrong weights. OBS also represents a major improvement over other methods, such as Optimal Brain Damage [Le Cun, Denker and Solla, 1990], because ours uses the full off-diagonal information of the Hessian matrix H. Crucial to OBS is a recursion relation for calculating H-1 from training data and structural information of the net. We illustrate OBS on standard benchmark problems — the MONK’s problems. The most successful method in a recent competition in machine learning [Thrun et al., 1991] was backpropagation using weight decay, which yielded a network with 58 weights for one MONK’s problem. OBS requires only 14 weights for the same performance accuracy. On two other MONK’s problems, our method required only 38% and 10% of the weights found by magnitude-based pruning. INTRODUCTION A central problem in machine learning is minimizing the system complexity (description length, VC-dimension, etc.) consistent with the training data. In neural networks this problem is usually cast as minimizing the number of connection weights. It is well-known that without such elimination of weights overfitting problems and thus poor performance on untrained

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Learning to Prune Deep Neural Networks via Layer-wise Optimal Brain Surgeon

How to develop slim and accurate deep neural networks has become crucial for realworld applications, especially for those employed in embedded systems. Though previous work along this research line has shown some promising results, most existing methods either fail to significantly compress a well-trained deep network or require a heavy retraining process for the pruned deep network to re-boost...

متن کامل

Comparing Adaptive and Non - AdaptiveConnection Pruning With Pure Early

|Neural network pruning methods on the level of individual network parameters (e.g. connection weights) can improve generalization, as is shown in this empirical study. However, an open problem in the pruning methods known today (OBD, OBS, autoprune, epsiprune) is the selection of the number of parameters to be removed in each pruning step (pruning strength). This work presents a pruning method...

متن کامل

Optimal Brain Surgeon: Extensions and performance comparisons

We extend Optimal Brain Surgeon (OBS) a second-order method for pruning networks to allow for general error measures, and explore a reduced computational and storage implementation via a dominant eigenspace decomposition. Simulations on nonlinear, noisy pattern classification problems reveal that OBS does lead to improved generalization, and performs favorably in comparison with Optimal Brain D...

متن کامل

An Analysis of Artificial Neural Network Pruning Algorithms in Relation to Land Cover Classification Accuracy

Artificial neural networks (ANNs) have been widely used for many classification purposes and generally proved to be more powerful than conventional statistical techniques. However, the use of ANNs requires decisions on the part of the user which may affect the accuracy of the resulting classification. One of these decisions concerns the determination of the optimum network structure for a parti...

متن کامل

A Comparative Study of Neural Network Optimization Techniques

In the last years we developed ENZO, an evolutionary neural network optimizer which surpasses other algorithms with regard to performance and scalability. In this study we compare ENZO to standard techniques for topology optimization: Optimal Brain Surgeon (OBS), Magnitude based Pruning (MbP), and to an improved algorithm deduced from OBS (unit-OBS). Furthermore we compare results to a newly pr...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1992